Distributional models of verb meaning: syntactic versus lexical contexts
نویسنده
چکیده
Over the last decade or so, distributional methods have become the mainstay of semantic modelling in Computational Linguistics. As such, they have also been applied the automatic modelling of verb meaning. However, more than with other lexical categories, the research into verb semantics has taken its inspiration from the idea that a verb's meaning is strongly linked to its syntactic behaviour and more specifically, to its selectional preferences. Depending on how they use these selectional preferences, distributional models of verb meaning come in two flavours. The first approach has its historical origins in the linguistic research tradition into verb valency and frame sem-natics and is in principle purely syntactical in nature. A verb's semantic category is said to be inferrable from its distribution over subcategorization (subcat) frames, i.e. the possible combinations of syntactic verb arguments like subject, direct object, indirect object etc. Additionally, this purely syntactic information can be extended with some high-level semantic information like the animacy of the verb arguments (see Schulte im Walde (2006) for an overview). Whereas this first, syntax-oriented approach is specifically geared towards verbs, the second approach is more generally applicable to all lexical categories and is a direct implementation of the ideas of Harris (1954). These so-called word space models use other words as context features with a specific implementation using only those context words that co-occur in a given dependency relation to the target word (see Padó and Lapata (2007) for an overview). In the first approach, one context feature is a possible combination of syntactic arguments that a verb can govern. In the second approach, one specific context feature corresponds to one lexeme plus its syntactic relation to the target verb. Whereas the first approach is mostly used to automatically induce Levin-style verb classes, the second approach is typically applied to retrieve semantic equivalents for specific verbs (but see Li and Brew (2008) for a comparison of the two methods on the task of inducing Levin-style classes). In this presentation we will try to have a closer look at the kind of semantic information that is captured by these two distinct types of distri-butional methods for verb meaning. For a sample of 1000 frequent Dutch
منابع مشابه
Distinguishing the Similar: An analysis of the semantic distinctions captured by distributional models of verb meaning
As in lexical semantics in general, distributional methods have also proven a successful technique for the automatic modeling of verb meaning. However, much more than with other lexical categories, the research into verb semantics has been based on the idea that a verb’s meaning is strongly linked to its syntactic behavior and more specifically, to its selectional preferences. This has led dist...
متن کاملInequality between the classes: Phonological and distributional typicality as predictors of lexical processing
Information about the syntactic category of a word can be derived from a number of complementary sources. We focus here on phonological and distributional cues for distinguishing nouns and verbs that have been proposed as useful for language acquisition. In this paper we assessed the extent to which this information affects lexical processing in adults. We hypothesised that the phonological or ...
متن کاملTowards a Representation of Verbal Semantics - An Approach Based on Near-synonyms
In this paper we propose using the distributional differences in the syntactic patterns of near-synonyms to deduce the relevant components of verb meaning. Our method involves determining the distributional differences in syntactic patterns, deducing the semantic features from the syntactic phenomena, and testing the semantic features in new syntactic frames. We determine the distributional dif...
متن کاملCategory-theoretic quantitative compositional distributional models of natural language semantics
This thesis is about the problem of compositionality in distributional semantics. Distributional semantics presupposes that the meanings of words are a function of their occurrences in textual contexts. It models words as distributions over these contexts and represents them as vectors in high dimensional spaces. The problem of compositionality for such models concerns itself with how to produc...
متن کاملDistributional learning and lexical category acquisition: What makes words easy to categorize?
In this study, results of computational simulations on English child-directed speech are presented to uncover what distributional properties of words make it easier to group them into lexical categories. This analysis provides evidence that words are easier to categorize when (i) they are hard to predict given the contexts they occur in; (ii) they occur in few different contexts; and (iii) thei...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
عنوان ژورنال:
دوره شماره
صفحات -
تاریخ انتشار 2010